Robustness of Bayesian Pool-Based Active Learning Against Prior Misspecification

نویسندگان

  • Viet Cuong Nguyen
  • Nan Ye
  • Wee Sun Lee
چکیده

We study the robustness of active learning (AL) algorithms against prior misspecification: whether an algorithm achieves similar performance using a perturbed prior as compared to using the true prior. In both the average and worst cases of the maximum coverage setting, we prove that all α-approximate algorithms are robust (i.e., near α-approximate) if the utility is Lipschitz continuous in the prior. We further show that robustness may not be achieved if the utility is non-Lipschitz. This suggests we should use a Lipschitz utility for AL if robustness is required. For the minimum cost setting, we can also obtain a robustness result for approximate AL algorithms. Our results imply that many commonly used AL algorithms are robust against perturbed priors. We then propose the use of a mixture prior to alleviate the problem of prior misspecification. We analyze the robustness of the uniform mixture prior and show experimentally that it performs reasonably well in practice.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Pool-Based Agnostic Experiment Design in Linear Regression

We address the problem of batch active learning (or experiment design) in regression scenarios, where the best input points to label is chosen from a ‘pool’ of unlabeled input samples. Existing active learning methods often assume that the model is correctly specified, i.e., the unknown learning target function is included in the model at hand. However, this assumption may not be fulfilled in p...

متن کامل

Imputation Estimators Partially Correct for Model Misspecification

Inference problems with incomplete observations often aim at estimating population properties of unobserved quantities. One simple way to accomplish this estimation is to impute the unobserved quantities of interest at the individual level and then take an empirical average of the imputed values. We show that this simple imputation estimator can provide partial protection against model misspeci...

متن کامل

Robust Calibration of Financial Models Using Bayesian Estimators

We consider a general calibration problem for derivative pricing models, which we reformulate into a Bayesian framework to attain posterior distributions for model parameters. It is then shown how the posterior distribution can be used to estimate prices for exotic options. We apply the procedure to a discrete local volatility model and work in great detail through numerical examples to clarify...

متن کامل

Qualitative Robustness in Bayesian Inference

We develop a framework for quantifying the sensitivity of the distribution of posterior distributions with respect to perturbations of the prior and data generating distributions in the limit when the number of data points grows towards infinity. In this generalization of Hampel [47] and Cuevas’ [18] notion of qualitative robustness to Bayesian inference, posterior distributions are analyzed as...

متن کامل

Improving importance estimation in pool-based batch active learning for approximate linear regression

Pool-based batch active learning is aimed at choosing training inputs from a 'pool' of test inputs so that the generalization error is minimized. P-ALICE (Pool-based Active Learning using Importance-weighted least-squares learning based on Conditional Expectation of the generalization error) is a state-of-the-art method that can cope with model misspecification by weighting training samples acc...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2016